Alpha-divergence minimization for deep Gaussian processes

نویسندگان

چکیده

This paper proposes the minimization of α-divergences for approximate inference in context deep Gaussian processes (DGPs). The proposed method can be considered as a generalization variational (VI) and expectation propagation (EP), two previously used methods DGPs. Both VI EP are based on Kullback-Leibler divergence. is scalable version power propagation, that introduces an extra parameter α specifies targeted α-divergence to optimized. In particular, such recover solution when α→0 α→1. An exhaustive experimental evaluation shows via feasible DGPs choosing intermediate values between 0 1 give better results some problems. means one improve training Importantly, allows stochastic optimization techniques, making it able address datasets with several millions instances.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Gaussian Processes

In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief network based on Gaussian process mappings. The data is modeled as the output of a multivariate GP. The inputs to that Gaussian process are then governed by another GP. A single layer model is equivalent to a standard GP or the GP latent variable model (GP-LVM). We perform inference in the model by approxim...

متن کامل

The Rate of Entropy for Gaussian Processes

In this paper, we show that in order to obtain the Tsallis entropy rate for stochastic processes, we can use the limit of conditional entropy, as it was done for the case of Shannon and Renyi entropy rates. Using that we can obtain Tsallis entropy rate for stationary Gaussian processes. Finally, we derive the relation between Renyi, Shannon and Tsallis entropy rates for stationary Gaussian proc...

متن کامل

Random Feature Expansions for Deep Gaussian Processes

The composition of multiple Gaussian Processes as a Deep Gaussian Process (DGP) enables a deep probabilistic nonparametric approach to flexibly tackle complex machine learning problems with sound quantification of uncertainty. Existing inference approaches for DGP models have limited scalability and are notoriously cumbersome to construct. In this work we introduce a novel formulation of DGPs b...

متن کامل

Contrastive Divergence In Gaussian Diffusion Processes

Contrastive Divergence (CD) is a recent learning rule found to work very well in practice despite its still unclear theoretical underpinnings (Carreira-Perpiñan and Hinton, 2005; Hinton, 2002; Hinton and Salakhutdinov, 2006; Hyvärinen, 2006; MacKay, 2001; Roth and Black, 2005; Williams and Agakov, 2002; Yuille, 2004). This paper presents a theoretical analysis of CD in Gaussian Diffusion Proces...

متن کامل

Deep Multi-fidelity Gaussian Processes

We develop a novel multi-fidelity framework that goes far beyond the classical AR(1) Co-kriging scheme of Kennedy and O’Hagan (2000). Our method can handle general discontinuous cross-correlations among systems with different levels of fidelity. A combination of multi-fidelity Gaussian Processes (AR(1) Co-kriging) and deep neural networks enables us to construct a method that is immune to disco...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: International Journal of Approximate Reasoning

سال: 2022

ISSN: ['1873-4731', '0888-613X']

DOI: https://doi.org/10.1016/j.ijar.2022.08.003